21 - Pattern Recognition [PR] - PR 17 [ID:23034]
50 von 87 angezeigt

Welcome back to pattern recognition. So today we want to talk a bit about norms and we will

look into different variants of vector norms, matrix norms and the like and have a couple

of examples how they are actually applied.

So let's look into norms and later on also norm dependent linear regression. So we've seen that

norms and similarity measures play an important role in machine learning and pattern recognition.

In this chapter we want to summarize important definitions and facts on norms and I think this

is also a very good refresher. If you're not so familiar with norms this video will be a very

good summary of the most important facts on norms. So we'll consider the problem of linear

regression then later in the next video for different norms but in this video we'll mainly

talk about norms and their properties. So later we also look briefly into the associated optimization

problems. So let's start with the inner product, the inner product of two vectors that we already

used in this lecture. We write this up as x transpose y and this essentially is the sum over

the element wise multiplication. This can then be used to define the Euclidean L2 norm. So this is

essentially the inner product of the vector with itself and then we take the square root. So we can

write this as x transpose x square root or you could say that's element wise the square summed

up and then taken the square root. Also we can define inner products of matrices and if we want

to do that then we see that the inner product of a matrix is essentially x transpose y and then we

take the trace of the resulting matrix and this then results in a sum over all the elements of x

multiplied with all the elements of y and then we compute the essentially two-dimensional sum

here because they're matrices but we mix all the elements with each other and sum them up. Now if

you follow this definition then you can find the Frobenius norm as the inner product of x with

itself so this means that we essentially square all of the elements and then sum them up. So this

is the Frobenius norm. So what else do we have? Well what's a norm actually? The function that we

indicate with these double bars is called a norm if it is non-negative so the result of the norm is

always zero or greater to zero for all given x then it should be the finite which means that it

will be zero only if all of the entries of x are zero and it should be homogeneous which means that

if you multiply a with x and a is a scalar value and you take the norm then this is the same as

taking the absolute value of a and multiplying it with the norm of x and this is of course true

for scalar values. Then it should also fulfill the triangle inequality which means that for all x and

y if you have some x and you add y and take the norm then this should be lesser or equal to the

norm of x plus the norm of y also known as the triangle inequality. Then we can also find other

norms for example the L0 norm which is essentially denoting the number of non-zero entries despite

its name it's not really a norm so we will see that and you see that often used also in rigid

the L0 norm and also other variants that are actually not norms but they're used in this

context. So generally this will allow us to define an LP norm and now P is a scalar value and note

that this is small caps P and it's defined as norm with P greater or equal to one of a d-dimensional

vector and then you can compute this essentially with the element wise x i's to the absolute value

to the power of P then summed up and then you take the Pth root of the respective sum. Technically

you can also do this with values lower than one but then you end up with constellations that no

longer fulfill all of the previous norm properties. In literature you can also find this as a 0 norm

or as L0.5 norm and so on. So you essentially use this same idea of computing them but you use values

that are smaller than one for this. Now let's look into real norms so there's the L1 norm which is

essentially the sum of absolute values then there's the L2 norm that is essentially the sum of squares

and the square root and then there's also other norms following this direction up to the L infinity

norm and this is also called maximum norm and here you essentially pick the maximum absolute value

that occurs in the vector and return it as the respective result of the norm. So there's also

LP norms but now P is a capital P and bold so this is a matrix and if it is a symmetric positive

definite matrix then the quadratic LP norm can be given in the following way so this is essentially

x transpose times P times x so it's essentially a weighted distance to itself weighted by P and

then you take the square root. So you can also write this up in the following way if you split

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

00:11:37 Min

Aufnahmedatum

2020-11-06

Hochgeladen am

2020-11-06 23:58:16

Sprache

en-US

In this video, we introduce norms on matrices and vectors and the corresponding unit balls.

This video is released under CC BY 4.0. Please feel free to share and reuse.

For reminders to watch the new video follow on Twitter or LinkedIn. Also, join our network for information about talks, videos, and job offers in our Facebook and LinkedIn Groups.

Music Reference: Damiano Baldoni - Thinking of You

Einbetten
Wordpress FAU Plugin
iFrame
Teilen